Efficient Algorithms for Learning to Play Repeated Games Against Computationally Bounded Adversaries

نویسندگان

  • Yoav Freund
  • Michael Kearns
  • Yishay Mansour
  • Dana Ron
  • Ronitt Rubinfeld
  • Robert E. Schapire
چکیده

In the game theory literature, there is an intriguing line of research on the problem of playing a repeated matrix game against an adversary whose computational resources are limited in some way. Perhaps the main way in which this research differs from classical game theory lies in the fact that when our adversary is not playing the minimax optimal strategy for the game, we may be able to attain payoff that is significantly greater than the minimax optimum. In this situation, the correct measure of our performance is in comparison to the optimum achievable against the particular adversary, not to the minimax optimum. The typical approach is to assume that the adversary’s strategy is a member of some natural class of computationally bounded strategies — most often, a class of finite automata. (For a survey on the area of “bounded rationality”, see the paper of Kalai [4].) Many previous papers examine how various aspects of classical game theory change in this setting; a good example is the question of whether cooperation is a stable solution for prisoner’s dilemma when both players are finite automata [6, 8]. Some authors have examined the further problem of learning to play optimally against an adversary whose precise strategy is unknown, but is constrained to lie in some known class of strategies (for instance, see Gilboa and Samet [3]). It is this research that forms our starting point. The previous work on learning to play optimally usually does not explicitly take into account the computational efficiency of the learning algorithm, and often gives algorithms whose running time is exponential in some natural measure of the adversary’s complexity; a notable recent exception is the work

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Eecient Algorithms for Learning to Play Repeated Games against Computationally Bounded Adversaries

We study the problem of eeciently learning to play a game optimally against an unknown adversary chosen from a computationally bounded class. We both contribute to the line of research on playing games against nite automata, and expand the scope of this research by considering new classes of adversaries. We introduce the natural notions of games against recent history adversaries (whose current...

متن کامل

4. Experimentation: the Effect of Exploration on the Agent's Performance

Model-based learning of interaction strategies in repeated-games has received a lot of attention in the game-theory literature. Gilboa & Samet [10] deal with bounded regular players. They describe a model-based learning strategy for repeated games that learns the best response against any regular strategy. Their procedure enumerates the set of all automata and chooses the current opponent model...

متن کامل

Convergence, Targeted Optimality, and Safety in Multiagent Learning

This paper introduces a novel multiagent learning algorithm, Convergence with Model Learning and Safety (or CMLeS in short), which achieves convergence, targeted optimality against memory-bounded adversaries, and safety, in arbitrary repeated games. The most novel aspect of CMLeS is the manner in which it guarantees (in a PAC sense) targeted optimality against memory-bounded adversaries, via ef...

متن کامل

Online Multiagent Learning against Memory Bounded Adversaries

The traditional agenda in Multiagent Learning (MAL) has been to develop learners that guarantee convergence to an equilibrium in self-play or that converge to playing the best response against an opponent using one of a fixed set of known targeted strategies. This paper introduces an algorithm called Learn or Exploit for Adversary Induced Markov Decision Process (LoE-AIM) that targets optimalit...

متن کامل

Designing Better Resource Allocation Strategy against Human Adversaries in Security Games

Stackelberg games have garnered significant attention in recent years given their deployment for real world security, such as ARMOR, IRIS and GUARDS. Most of these systems have adopted the standard game-theoretical assumption that adversaries are perfectly rational, which may not hold in real-world security problems due to the bounded rationality of human adversaries and could potentially reduc...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1995